Optimization of ℓp-regularized Linear Models via Coordinate Descent
نویسندگان
چکیده
منابع مشابه
Penalized Bregman Divergence Estimation via Coordinate Descent
Variable selection via penalized estimation is appealing for dimension reduction. For penalized linear regression, Efron, et al. (2004) introduced the LARS algorithm. Recently, the coordinate descent (CD) algorithm was developed by Friedman, et al. (2007) for penalized linear regression and penalized logistic regression and was shown to gain computational superiority. This paper explores...
متن کاملRegularization Paths for Generalized Linear Models via Coordinate Descent.
We develop fast algorithms for estimation of generalized linear models with convex penalties. The models include linear regression, two-class logistic regression, and multinomial regression problems while the penalties include ℓ(1) (the lasso), ℓ(2) (ridge regression) and mixtures of the two (the elastic net). The algorithms use cyclical coordinate descent, computed along a regularization path....
متن کاملSparse Linear Programming via Primal and Dual Augmented Coordinate Descent
Over the past decades, Linear Programming (LP) has been widely used in differentareas and considered as one of the mature technologies in numerical optimization.However, the complexity offered by state-of-the-art algorithms (i.e. interior-pointmethod and primal, dual simplex methods) is still unsatisfactory for problems inmachine learning with huge number of variables and constr...
متن کاملSolving norm constrained portfolio optimization via coordinate-wise descent algorithms
In this paper we demonstrate that coordinate-wise descent algorithms can be used to solve portfolio selection problems in which asset weights are constrained by l q norms for 1 ≤ q ≤ 2. A special case of the such problems is when q = 1. The l 1 norm constraint promotes zero values for the weight vector, leading to an automatic asset selection for the portfolio. We first consider the case of min...
متن کاملLasso Regularization Paths for NARMAX Models via Coordinate Descent
We propose a new algorithm for estimating NARMAX models with L1 regularization for models represented as a linear combination of basis functions. Due to the L1-norm penalty the Lasso estimation tends to produce some coefficients that are exactly zero and hence gives interpretable models. The novelty of the contribution is the inclusion of error regressors in the Lasso estimation (which yields a...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Schedae Informaticae
سال: 2017
ISSN: 2083-8476
DOI: 10.4467/20838476si.16.005.6186